Orthogonal polynomials

In mathematics, an orthogonal polynomial sequence is an infinite sequence of real polynomials

p_0,\ p_1,\ p_2,\ \ldots

of one variable x, in which each pn has degree n, and such that any two different polynomials in the sequence are orthogonal to each other under a particular version of the L2 inner product.

The field of orthogonal polynomials developed in the late 19th century from a study of continued fractions by P. L. Chebyshev and was pursued by A.A. Markov and T.J. Stieltjes and by a few other mathematicians. Since then, applications have been developed in many areas of mathematics and physics.

Contents

Definition

The theory of orthogonal polynomials includes many definitions of orthogonality. In abstract notation, is convenient to write

\langle p,q\rangle=0

when the polynomials p(x) and q(x) are orthogonal. A sequence of orthogonal polynomials, then, is a sequence of polynomials

p_0,\ p_1,\ p_2,\ \ldots

such that pn has degree n and all distinct members of the sequence are orthogonal to each other.

The algebraic and analytic properties of the polynomials depend upon the specific assumptions about the operator \langle \cdot,\cdot\rangle. In the classical formulation, the operator is defined in terms of the integral of a weighted product (see below) and happens to be an inner product. Other formulations remove various assumptions, for example in the context of Hilbert spaces or non-Hermitian operators (see below). Most of the discussion in this article applies to the classical definition.

Classical formulation

Let [x_1, x_2] be an interval in the real line (where x_1 = -\infty and x_2 = \infty are allowed). This is called the interval of orthogonality. Let

W�: [x_1, x_2] \to \mathbb{R}

be a function on the interval, that is strictly positive on the interior (x_1, x_2), but which may be zero or go to infinity at the end points. Additionally, W must satisfy the requirement that, for any polynomial f, the integral

\int_{x_1}^{x_2} f(x) W(x) \; dx

is finite. Such a W is called a weight function.

Given any x_1, x_2, and W as above, define an operation on pairs of polynomials f and g by

\langle f, g \rangle = \int_{x_1}^{x_2} f(x) g(x) W(x) \; dx.

This operation is an inner product on the vector space of all polynomials. It induces a notion of orthogonality in the usual way, namely that two polynomials are orthogonal if their inner product is zero.

Generalizations

Many alternative theories of orthogonal polynomials have been studied and, along with the classical theory, remain active areas of research.[1] Some aspects of the classical theory generalize when certain assumptions are lifted, and new properties can arise in different contexts.

In some theories the polynomials may act on other algebraic objects such as the complex numbers, matrices, and the unit circle (as a subset of the complex numbers).

Much of the general theory is for operators \langle\cdot,\cdot\rangle that satisfy the axioms of an inner product. This includes inner products within a Hilbert space (where the polynomials can be interpreted as an orthogonal basis) and inner products that can be defined as an integral of the form

\langle p,q\rangle = \int p(x)q(x) d\mu(x)

where μ is a positive measure; this in turn includes the classical definition as well as the probabilistic definition (where the measure is a probability measure) and the discrete definition (where the integral is an infinite weighted sum).

The effects of lifting the inner product assumption of positive definiteness have also been studied (e.g. negative weights, discrete coefficients or non-Hermitian operators). In this theory, the terms system and sequence of orthogonal polynomials are distinct because pairs of polynomials of the same degree may be orthogonal.

For the remainder of this article the classical definition is assumed.

Standardization

The chosen inner product induces a norm on polynomials in the usual way:

|| f || = \langle f, f \rangle^{1 / 2}.

When making an orthogonal basis, one may be tempted to make an orthonormal basis, that is, one in which all basis elements have norm 1. For polynomials, this would often result in simple square roots in the coefficients. Instead, polynomials are often scaled in a way that mathematicians agree on, that makes the coefficients and other formulas simpler. This is called standardization. The "classical" polynomials listed below have been standardized, typically by setting their leading coefficients to some specific quantity, or by setting a specific value for the polynomial. This standardization has no mathematical significance; it is just a convention. Standardization also involves scaling the weight function in an agreed-upon way.

Denote by h_n the square of the norm of p_n:

h_n = \langle p_n, p_n \rangle.

The values of h_n for the standardized classical polynomials are listed in the table below. In this notation,

\langle p_m, p_n \rangle = \delta_{mn} \sqrt{h_m h_n},

where δmn is the Kronecker delta.

Example: Legendre polynomials

The simplest classical orthogonal polynomials are the Legendre polynomials, for which the interval of orthogonality is [−1, 1] and the weight function is simply 1:

P_0(x) = 1,\,
P_1(x) = x,\,
P_2(x) = \frac{3x^2-1}{2},\,
P_3(x) = \frac{5x^3-3x}{2},\,
P_4(x) = \frac{35x^4-30x^2+3}{8},\,
\vdots

These are all orthogonal over [−1, 1]; whenever m ≠ n,

\int_{-1}^1 P_m(x) P_n(x) \, dx = 0.

The Legendre polynomials are standardized so that Pn(1) = 1 for all n.

Non-classical example

The simplest non-classical orthogonal polynomials are the monomials

p_n(x) = x^n\,

for n ≥ 0 which are orthogonal under the inner product defined by

\langle x^m, x^n\rangle = \delta_{mn}.

This inner product cannot be defined in the classical sense as a weighted integral of a product, or even as a measure of a product (otherwise 0=\langle 1,x^2\rangle=\langle x,x\rangle=1). However the monomials are orthogonal on the unit circle as a subset of the complex numbers, using the path integral

\langle p, q\rangle = \frac{1}{2\pi i}\int_{|z|=1}\frac{1}{z}p(z)\overline{q(z)} \, dz = \frac{1}{2\pi}\int_0^{2\pi}p(e^{i\theta})\overline{q(e^{i\theta})} \, d\theta

where \overline{z} is the complex conjugate of z. The properties of orthogonal polynomials on the unit circle differ from those of classical orthogonal polynomials (such as the form of the recurrence relations and the distribution of roots) and are related to the theory of Fourier series.

General properties of orthogonal polynomial sequences

All orthogonal polynomial sequences have a number of elegant and fascinating properties. Before proceeding with them:

Lemma 1: Given an orthogonal polynomial sequence p_i(x), any nth-degree polynomial S(x) can be expanded in terms of p_0, \dots, p_n. That is, there are coefficients \alpha_0, \dots, \alpha_n such that

S(x)=\sum_{i=0}^n \alpha_i p_i(x).

Proof by mathematical induction. Choose \alpha_n so that the x^n term of S(x) matches that of \alpha_n P_n(x). Then S(x)-\alpha_n P_n(x) is an (n − 1)th-degree polynomial. Continue downward.

The coefficients \alpha_i can be calculated directly using orthogonality. First multiply S by p_k and weight function W, then integrate:

\int S(x) p_k(x) W(x) \, dx = \sum_{i=0}^n {\alpha}_i \int p_i(x) p_k(x) W(x) \, dx = \alpha_k \int p_k^2(x) W(x) \, dx,

giving

\alpha_k  = {\int S(x) p_k(x) W(x) \, dx \over\int p_k^2(x) W(x) \, dx}.

Lemma 2: Given an orthogonal polynomial sequence, each of its polynomials is orthogonal to any polynomial of strictly lower degree.

Proof: Given n, any polynomial of degree n − 1 or lower can be expanded in terms of p_0, \dots, p_{n-1}. Polynomial p_n\, is orthogonal to each of them.

Minimal norm

Each polynomial in an orthogonal sequence has minimal norm among all polynomials with the same degree and leading coefficient.

An interpretation of this result is that orthogonal polynomials are minimal in a generalized least squares sense. For example, the classical orthogonal polynomials have a minimal weighted mean square value.

Recurrence relations

Any orthogonal sequence has a recurrence formula relating any three consecutive polynomials in the sequence:

p_{n+1}\ =\ (a_nx+b_n)\ p_n\ -\ c_n\ p_{n-1}.

The coefficients a, b, and c depend on n, as well as the standardization.

The values of a_n, b_n and c_n can be worked out directly. Let k_j and k_j' be the first and second coefficients of p_j:

p_j(x)=k_jx^j+k_j'x^{j-1}+\cdots

and h_j be the inner product of p_j with itself:

h_j\ =\ \langle p_j,\ p_j \rangle.

We have

a_n=\frac{k_{n+1}}{k_n},\qquad b_n=a_n \left(\frac{k_{n+1}'}{k_{n+1}} -
\frac{k_n'}{k_n} \right), \qquad c_n=a_n \left(\frac{k_{n-1}h_n}{k_n h_{n-1}} \right).

Existence of real roots

Each polynomial in an orthogonal sequence has all n of its roots real, distinct, and strictly inside the interval of orthogonality.

Interlacing of roots

The roots of each polynomial lie strictly between the roots of the next higher polynomial in the sequence.

Differential equations leading to orthogonal polynomials

A very important class of orthogonal polynomials arises from a differential equation of the form

{Q(x)}\,f'' + {L(x)}\,f' + {\lambda}f = 0\,

where Q is a given quadratic (at most) polynomial, and L is a given linear polynomial. The function f, and the constant λ, are to be found.

(Note that it makes sense for such an equation to have a polynomial solution.
Each term in the equation is a polynomial, and the degrees are consistent.)

This is a Sturm-Liouville type of equation. Such equations generally have singularities in their solution functions f except for particular values of λ. They can be thought of a eigenvector/eigenvalue problems: Letting D be the differential operator, D(f) = Q f'' + L f'\,, and changing the sign of λ, the problem is to find the eigenvectors (eigenfunctions) f, and the corresponding eigenvalues λ, such that f does not have singularities and D(f) = λf.

The solutions of this differential equation have singularities unless λ takes on specific values. There is a series of numbers {\lambda}_0, {\lambda}_1, {\lambda}_2, \dots\, that lead to a series of polynomial solutions P_0, P_1, P_2, \dots\, if one of the following sets of conditions are met:

  1. Q is actually quadratic, L is linear, Q has two distinct real roots, the root of L lies strictly between the roots of Q, and the leading terms of Q and L have the same sign.
  2. Q is not actually quadratic, but is linear, L is linear, the roots of Q and L are different, and the leading terms of Q and L have the same sign if the root of L is less than the root of Q, or vice-versa.
  3. Q is just a nonzero constant, L is linear, and the leading term of L has the opposite sign of Q.

These three cases lead to the Jacobi-like, Laguerre-like, and Hermite-like polynomials, respectively.

In each of these three cases, we have the following:

Because of the constant of integration, the quantity R(x) is determined only up to an arbitrary positive multiplicative constant. It will be used only in homogeneous differential equations (where this doesn't matter) and in the definition of the weight function (which can also be indeterminate.) The tables below will give the "official" values of R(x) and W(x).

Rodrigues' formula

Under the assumptions of the preceding section, Pn(x) is proportional to \frac{1}{W(x)} \  \frac{d^n}{dx^n}\left(W(x)[Q(x)]^n\right).

This is known as Rodrigues' formula, after Olinde Rodrigues. It is often written

P_n(x) = \frac{1}{{e_n}W(x)} \  \frac{d^n}{dx^n}\left(W(x)[Q(x)]^n\right)

where the numbers en depend on the standardization. The standard values of en will be given in the tables below.

The numbers λn

Under the assumptions of the preceding section, we have

{\lambda}_n = - n \left( \frac{n-1}{2} Q'' + L' \right).

(Since Q is quadratic and L is linear, Q'' and L' are constants, so these are just numbers.)

Second form for the differential equation

Let R(x) = e^{\int \frac{L(x)}{Q(x)}\,dx}\,.

Then

(Ry')' = R\,y'' + R'\,y' = R\,y'' + \frac{R\,L}{Q}\,y'.

Now multiply the differential equation

{Q}\,y'' + {L}\,y' + {\lambda}\,y = 0\,

by R/Q, getting

R\,y'' + \frac{R\,L}{Q}\,y' + \frac{R\,\lambda}{Q}\,y = 0\,

or

(Ry')' + \frac{R\,\lambda}{Q}\,y = 0.\,

This is the standard Sturm-Liouville form for the equation.

Third form for the differential equation

Let S(x) = \sqrt{R(x)} = e^{\int \frac{L(x)}{2\,Q(x)}\,dx}.\,

Then

S' = \frac{S\,L}{2\,Q}.

Now multiply the differential equation

{Q}\,y'' + {L}\,y' + {\lambda}\,y = 0\,

by S/Q, getting

S\,y'' + \frac{S\,L}{Q}\,y' + \frac{S\,\lambda}{Q}\,y = 0\,

or

S\,y'' + 2\,S'\,y' + \frac{S\,\lambda}{Q}\,y = 0\,

But (S\,y)'' = S\,y'' + 2\,S'\,y' + S''\,y, so

(S\,y)'' + \left(\frac{S\,\lambda}{Q} - S''\right)\,y = 0,\,

or, letting u = Sy,

u'' + \left(\frac{\lambda}{Q} - \frac{S''}{S}\right)\,u = 0.\,

Formulas involving derivatives

Under the assumptions of the preceding section, let P_n^{[r]} denote the rth derivative of P_n. (We put the "r" in brackets to avoid confusion with an exponent.) P_n^{[r]} is a polynomial of degree n − r. Then we have the following:

There are also some mixed recurrences. In each of these, the numbers a, b, and c depend on n and r, and are unrelated in the various formulas.

There are an enormous number of other formulas involving orthogonal polynomials in various ways. Here is a tiny sample of them, relating to the Chebyshev, associated Laguerre, and Hermite polynomials:

Orthogonality

The differential equation for a particular λ may be written (omitting explicit dependence on x)

Q\ddot{f}_n+L\dot{f}_n+\lambda_nf_n=0

multiplying by (R/Q)f_m yields

Rf_m\ddot{f}_n+\textstyle\frac{R}{Q}Lf_m\dot{f}_n+\textstyle\frac{R}{Q}\lambda_nf_mf_n=0

and reversing the subscripts yields

Rf_n\ddot{f}_m+\textstyle\frac{R}{Q}Lf_n\dot{f}_m+\textstyle\frac{R}{Q}\lambda_mf_nf_m=0

subtracting and integrating:


\int_a^b \left[R(f_m\ddot{f}_n-f_n\ddot{f}_m)+
\textstyle\frac{R}{Q}L(f_m\dot{f}_n-f_n\dot{f}_m)\right] \, dx
+(\lambda_n-\lambda_m)\int_a^b \textstyle\frac{R}{Q}f_mf_n \, dx = 0

but it can be seen that


\frac{d}{dx}\left[R(f_m\dot{f}_n-f_n\dot{f}_m)\right]=
R(f_m\ddot{f}_n-f_n\ddot{f}_m)\,\,+\,\,R\textstyle\frac{L}{Q}(f_m\dot{f}_n-f_n\dot{f}_m)

so that:

\left[R(f_m\dot{f}_n-f_n\dot{f}_m)\right]_a^b\,\,+\,\,(\lambda_n-\lambda_m)\int_a^b \textstyle\frac{R}{Q}f_mf_n \, dx=0

If the polynomials f are such that the term on the left is zero, and \lambda_m \ne \lambda_n for m \ne n, then the orthogonality relationship will hold:

\int_a^b \textstyle\frac{R}{Q}f_mf_n \, dx=0

for m \ne n.

The classical orthogonal polynomials

The class of polynomials arising from the differential equation described above have many important applications in such areas as mathematical physics, interpolation theory, the theory of random matrices, computer approximations, and many others. All of these polynomial sequences are equivalent, under scaling and/or shifting of the domain, and standardizing of the polynomials, to more restricted classes. Those restricted classes are the "classical orthogonal polynomials".

Because all polynomial sequences arising from a differential equation in the manner described above are trivially equivalent to the classical polynomials, the actual classical polynomials are always used.

Jacobi polynomials

The Jacobi-like polynomials, once they have had their domain shifted and scaled so that the interval of orthogonality is [−1, 1], still have two parameters to be determined. They are \alpha and \beta in the Jacobi polynomials, written P_n^{(\alpha, \beta)}. We have Q(x) = 1-x^2\, and L(x) = \beta-\alpha-(\alpha+\beta+2)\, x. Both \alpha and \beta are required to be greater than −1. (This puts the root of L inside the interval of orthogonality.)

When \alpha and \beta are not equal, these polynomials are not symmetrical about x = 0.

The differential equation

(1-x^2)\,y'' + (\beta-\alpha-[\alpha+\beta+2]\,x)\,y' + {\lambda}\,y = 0\qquad \mathrm{with}\qquad\lambda = n(n+1+\alpha+\beta)\,

is Jacobi's equation.

For further details, see Jacobi polynomials.

Gegenbauer polynomials

When one sets the parameters \alpha and \beta in the Jacobi polynomials equal to each other, one obtains the Gegenbauer or ultraspherical polynomials. They are written C_n^{(\alpha)}, and defined as

C_n^{(\alpha)}(x) = \frac{\Gamma(2\alpha\!+\!n)\,\Gamma(\alpha\!+\!1/2)}
{\Gamma(2\alpha)\,\Gamma(\alpha\!+\!n\!+\!1/2)}\! \  P_n^{(\alpha-1/2, \alpha-1/2)}.

We have Q(x) = 1-x^2\, and L(x) = -(2\alpha+1)\, x. \alpha\, is required to be greater than −1/2.

(Incidentally, the standardization given in the table below would make no sense for α = 0 and n ≠ 0, because it would set the polynomials to zero. In that case, the accepted standardization sets C_n^{(0)}(1) = \frac{2}{n} instead of the value given in the table.)

Ignoring the above considerations, the parameter \alpha is closely related to the derivatives of C_n^{(\alpha)}:

C_n^{(\alpha+1)}(x) = \frac{1}{2\alpha}\! \  \frac{d}{dx}C_{n+1}^{(\alpha)}(x)

or, more generally:

C_n^{(\alpha+m)}(x) = \frac{\Gamma(\alpha)}{2^m\Gamma(\alpha+m)}\! \  C_{n+m}^{(\alpha)[m]}(x).

All the other classical Jacobi-like polynomials (Legendre, etc.) are special cases of the Gegenbauer polynomials, obtained by choosing a value of \alpha and choosing a standardization.

For further details, see Gegenbauer polynomials.

Legendre polynomials

The differential equation is

(1-x^2)\,y'' - 2x\,y' + {\lambda}\,y = 0\qquad \mathrm{with}\qquad\lambda = n(n+1).\,

This is Legendre's equation.

The second form of the differential equation is:

([1-x^2]\,y')' + \lambda\,y = 0.\,

The recurrence relation is

(n+1)\,P_{n+1}(x) = (2n+1)x\,P_n(x) - n\,P_{n-1}(x).\,

A mixed recurrence is

P_{n+1}^{[r+1]}(x) = P_{n-1}^{[r+1]}(x) + (2n+1)\,P_n^{[r]}(x).\,

Rodrigues' formula is

P_n(x) = \,\frac{1}{2^n\,n!} \  \frac{d^n}{dx^n}\left([x^2-1]^n\right).

For further details, see Legendre polynomials.

Associated Legendre polynomials

The Associated Legendre polynomials, denoted P_\ell^{(m)}(x) where \ell and m are integers with 0 \leqslant m  \leqslant \ell, are defined as

P_\ell^{(m)}(x) = (-1)^m\,(1-x^2)^{m/2}\ P_\ell^{[m]}(x).\,

The m in parentheses (to avoid confusion with an exponent) is a parameter. The m in brackets denotes the mth derivative of the Legendre polynomial.

These "polynomials" are misnamed -- they are not polynomials when m is odd.

They have a recurrence relation:

(\ell+1-m)\,P_{\ell+1}^{(m)}(x) = (2\ell+1)x\,P_\ell^{(m)}(x) - (\ell+m)\,P_{\ell-1}^{(m)}(x).\,

For fixed m, the sequence P_m^{(m)}, P_{m+1}^{(m)}, P_{m+2}^{(m)}, \dots are orthogonal over [−1, 1], with weight 1.

For given m, P_\ell^{(m)}(x) are the solutions of

(1-x^2)\,y'' -2xy' + \left[\lambda - \frac{m^2}{1-x^2}\right]\,y = 0\qquad \mathrm{with}\qquad\lambda = \ell(\ell+1).\,

Chebyshev polynomials

The differential equation is

(1-x^2)\,y'' - x\,y' + {\lambda}\,y = 0\qquad \mathrm{with}\qquad\lambda = n^2.\,

This is Chebyshev's equation.

The recurrence relation is

T_{n+1}(x) = 2x\,T_n(x) - T_{n-1}(x).\,

Rodrigues' formula is

T_n(x) = \frac{\Gamma(1/2)\sqrt{1-x^2}}{(-2)^n\,\Gamma(n+1/2)} \  \frac{d^n}{dx^n}\left([1-x^2]^{n-1/2}\right).

These polynomials have the property that, in the interval of orthogonality,

T_n(x) = \cos(n\,\arccos(x)).

(To prove it, use the recurrence formula.)

This means that all their local minima and maxima have values of −1 and +1, that is, the polynomials are "level". Because of this, expansion of functions in terms of Chebyshev polynomials is sometimes used for polynomial approximations in computer math libraries.

Some authors use versions of these polynomials that have been shifted so that the interval of orthogonality is [0, 1] or [−2, 2].

There are also Chebyshev polynomials of the second kind, denoted U_n\,

We have:

U_n = \frac{1}{n+1}\,T_{n+1}'.\,

For further details, including the expressions for the first few polynomials, see Chebyshev polynomials.

Laguerre polynomials

The most general Laguerre-like polynomials, after the domain has been shifted and scaled, are the Associated Laguerre polynomials (also called Generalized Laguerre polynomials), denoted L_n^{(\alpha)}. There is a parameter \alpha, which can be any real number strictly greater than −1. The parameter is put in parentheses to avoid confusion with an exponent. The plain Laguerre polynomials are simply the \alpha = 0 version of these:

L_n(x) = L_n^{(0)}(x).\,

The differential equation is

x\,y'' + (\alpha + 1-x)\,y' + {\lambda}\,y = 0\qquad \mathrm{with}\qquad\lambda = n.\,

This is Laguerre's equation.

The second form of the differential equation is

(x^{\alpha+1}\,e^{-x}\, y')' + {\lambda}\,x^{\alpha}\,e^{-x}\,y = 0.\,

The recurrence relation is

(n+1)\,L_{n+1}^{(\alpha)}(x) = (2n+1+\alpha-x)\,L_n^{(\alpha)}(x) - (n+\alpha)\,L_{n-1}^{(\alpha)}(x).\,

Rodrigues' formula is

L_n^{(\alpha)}(x) = \frac{x^{-\alpha}e^x}{n!} \  \frac{d^n}{dx^n}\left(x^{n+\alpha}\,e^{-x}\right).

The parameter \alpha is closely related to the derivatives of L_n^{(\alpha)}:

L_n^{(\alpha+1)}(x) = - \frac{d}{dx}L_{n+1}^{(\alpha)}(x)

or, more generally:

L_n^{(\alpha+m)}(x) = (-1)^m L_{n+m}^{(\alpha)[m]}(x).

Laguerre's equation can be manipulated into a form that is more useful in applications:

u = x^{\frac{\alpha-1}{2}}e^{-x/2}L_n^{(\alpha)}(x)

is a solution of

u'' + \frac{2}{x}\,u' + \left[\frac{\lambda}{x} - \frac{1}{4} - \frac{\alpha^2-1}{4x^2}\right]\,u = 0\qquad \mathrm{with}\qquad\lambda = n+\frac{\alpha+1}{2}.\,

This can be further manipulated. When \ell = \frac{\alpha-1}{2} is an integer, and n{\ge}\ell+1:

u = x^{\ell}e^{-x/2}L_{n-\ell-1}^{(2\ell+1)}(x)

is a solution of

u'' + \frac{2}{x}\,u' + \left[\frac{\lambda}{x} - \frac{1}{4} - \frac{\ell(\ell+1)}{x^2}\right]\,u = 0\text{ with }\lambda = n.\,

The solution is often expressed in terms of derivatives instead of associated Laguerre polynomials:

u = x^{\ell}e^{-x/2}L_{n+\ell}^{[2\ell+1]}(x).

This equation arises in quantum mechanics, in the radial part of the solution of the Schrödinger equation for a one-electron atom.

Physicists often use a definition for the Laguerre polynomials that is larger, by a factor of (n!), than the definition used here.

For further details, including the expressions for the first few polynomials, see Laguerre polynomials.

Hermite polynomials

The differential equation is

y'' - 2xy' + {\lambda}\,y = 0,\qquad \mathrm{with}\qquad\lambda = 2n.\,

This is Hermite's equation.

The second form of the differential equation is

(e^{-x^2}\,y')' + e^{-x^2}\,\lambda\,y = 0.\,

The third form is

(e^{-x^2/2}\,y)'' + ({\lambda}+1-x^2)(e^{-x^2/2}\,y) = 0.\,

The recurrence relation is

H_{n+1}(x) = 2x\,H_n(x) - 2n\,H_{n-1}(x).\,

Rodrigues' formula is

H_n(x) = (-1)^n\,e^{x^2} \  \frac{d^n}{dx^n}\left(e^{-x^2}\right).

The first few Hermite polynomials are

H_0(x) = 1\,
H_1(x) = 2x\,
H_2(x) = 4x^2-2\,
H_3(x) = 8x^3-12x\,
H_4(x) = 16x^4-48x^2+12\,

One can define the associated Hermite functions

{\psi}_n(x) = (h_n)^{-1/2}\,e^{-x^2/2}H_n(x).\,

Because the multiplier is proportional to the square root of the weight function, these functions are orthogonal over (-\infty, \infty) with no weight function.

The third form of the differential equation above, for the associated Hermite functions, is

\psi'' + ({\lambda}+1-x^2)\psi = 0.\,

The associated Hermite functions arise in many areas of mathematics and physics. In quantum mechanics, they are the solutions of Schrödinger's equation for the harmonic oscillator. They are also eigenfunctions (with eigenvalue (−i)n) of the continuous Fourier transform.

Many authors, particularly probabilists, use an alternate definition of the Hermite polynomials, with a weight function of e^{-x^2/2} instead of e^{-x^2}. If the notation He is used for these Hermite polynomials, and H for those above, then these may be characterized by

He_n(x) = 2^{-n/2}\,H_n\left(\frac{x}{\sqrt{2}}\right).

For further details, see Hermite polynomials.

Constructing orthogonal polynomials by the Gram–Schmidt process

The Gram–Schmidt process is an algorithm originally taken from linear algebra which removes linear dependency from a set of given vectors in an inner product space. The inner product as defined on all polynomials allows us to apply the Gram–Schmidt process to an arbitrary set of polynomials. The process removes linear dependencies from the polynomials, yielding sets of orthogonal polynomials. Given various initial polynomial sequences and weighting functions, different orthogonal polynomial sequences can be produced.

We define a projection operator on the polynomials as:

\mathrm{proj}_{f}\,(g) = {\langle f, g\rangle\over\langle f, f\rangle}f = { \int\limits_{x_1}^{x_2} f(x) g(x) W(x) \; dx \over  \int\limits_{x_1}^{x_2} (f(x))^2 W(x) \; dx} f(x).

To apply the algorithm, we define our set of original polynomials g_1, g_2, \ldots, g_k and generate a sequence of orthogonal polynomials f_1, f_2, \ldots, f_k using:


\begin{align}
f_1 & = g_1, \\
f_2 & = g_2-\mathrm{proj}_{f_1}\,(g_2), \\
f_3 & = g_3-\mathrm{proj}_{f_1}\,(g_3)-\mathrm{proj}_{f_2}\,(g_3), \\
& {}\  \  \vdots \\
f_k & = g_k-\sum_{j=1}^{k-1}\mathrm{proj}_{f_j}\,(g_k).
\end{align}

If an orthonormal sequence is required, a polynomial normalization operation can be defined as:

\hat{f} = {{f}\over{|| f ||}} = {{f(x)}\over{\sqrt{\int\limits_{x_1}^{x_2} (f(x))^2 W(x) \; dx}}}.

Care must be taken if the process is implemented on computer as the Gram–Schmidt process is numerically unstable. However, as many computational platforms implement rational numbers with arbitrary-precision arithmetic the problem can often be easily avoided.

Constructing orthogonal polynomials by using moments

Let

 \mu_n = \int_\mathbb{R} x^n\,d\mu

be the moments of a measure μ. Then the polynomial sequence defined by

 p_n(x) = \det\left[ 
\begin{matrix}
\mu_0 & \mu_1 & \mu_2 & \cdots & \mu_n \\
\mu_1 & \mu_2 & \mu_3 & \cdots & \mu_{n+1} \\
\mu_2 & \mu_3 & \mu_4 & \cdots & \mu_{n+2} \\
\vdots & \vdots & \vdots &      & \vdots \\
\mu_{n-1} & \mu_n & \mu_{n+1} & \cdots & \mu_{2n-1} \\
1 & x & x^2 & \cdots & x^n
\end{matrix}
\right]

is a sequence of orthogonal polynomials with respect to the measure μ. To see this, consider the inner product of pn(x) with xk for any k < n. We will see that the value of this inner product is zero[2].


\begin{align}
\int_\mathbb{R} x^k p_n(x)\,d\mu
& {} = \int_\mathbb{R} x^k \det\left[ 
\begin{matrix}
\mu_0 & \mu_1 & \mu_2 & \cdots & \mu_n \\
\mu_1 & \mu_2 & \mu_3 & \cdots & \mu_{n+1} \\
\mu_2 & \mu_3 & \mu_4 & \cdots & \mu_{n+2} \\
\vdots & \vdots & \vdots &      & \vdots \\
\mu_{n-1} & \mu_n & \mu_{n+1} & \cdots & \mu_{2n-1} \\
1 & x & x^2 & \cdots & x^n
\end{matrix} \right]
\,d\mu \\  \\
& {} = \int_\mathbb{R} \det\left[ 
\begin{matrix}
\mu_0 & \mu_1 & \mu_2 & \cdots & \mu_n \\
\mu_1 & \mu_2 & \mu_3 & \cdots & \mu_{n+1} \\
\mu_2 & \mu_3 & \mu_4 & \cdots & \mu_{n+2} \\
\vdots & \vdots & \vdots &      & \vdots \\
\mu_{n-1} & \mu_n & \mu_{n+1} & \cdots & \mu_{2n-1} \\
x^k & x^{k+1} & x^{k+2} & \cdots & x^{k+n}
\end{matrix} \right]
\,d\mu \\  \\
& {} = \det\left[ 
\begin{matrix}
\mu_0 & \mu_1 & \mu_2 & \cdots & \mu_n \\
\mu_1 & \mu_2 & \mu_3 & \cdots & \mu_{n+1} \\
\mu_2 & \mu_3 & \mu_4 & \cdots & \mu_{n+2} \\
\vdots & \vdots & \vdots &      & \vdots \\
\mu_{n-1} & \mu_n & \mu_{n+1} & \cdots & \mu_{2n-1} \\
\displaystyle \int_\mathbb{R} x^k \, d\mu & \displaystyle \int_\mathbb{R} x^{k+1} \, d\mu & \displaystyle \int_\mathbb{R} x^{k+2} \, d\mu & \cdots & \displaystyle \int_\mathbb{R} x^{k+n} \, d\mu
\end{matrix} \right] \\  \\
& {} = \det \left[ 
\begin{matrix}
\mu_0 & \mu_1 & \mu_2 & \cdots & \mu_n \\
\mu_1 & \mu_2 & \mu_3 & \cdots & \mu_{n+1} \\
\mu_2 & \mu_3 & \mu_4 & \cdots & \mu_{n+2} \\
\vdots & \vdots & \vdots &      & \vdots \\
\mu_{n-1} & \mu_n & \mu_{n+1} & \cdots & \mu_{2n-1} \\
\mu_k & \mu_{k+1} & \mu_{k+2} & \cdots & \mu_{k+n}
\end{matrix} \right] \\  \\
& {} = 0\text{ if } k < n,\text{ since the matrix has two identical rows}.
\end{align}

(The entry-by-entry integration merely says the integral of a linear combination of functions is the same linear combination of the separate integrals. It is a linear combination because only one row contains non-scalar entries.)

Thus pn(x) is orthogonal to xk for all k < n. That means this is a sequence of orthogonal polynomials for the measure μ.

Table of classical orthogonal polynomials

Name, and conventional symbol Chebyshev, \ T_n Chebyshev
(second kind), \ U_n
Legendre, \ P_n Hermite, \ H_n
Limits of orthogonality -1, 1\, -1, 1\, -1, 1\, -\infty, \infty
Weight, W(x)\, (1-x^2)^{-1/2}\, (1-x^2)^{1/2}\, 1\, e^{-x^2}
Standardization T_n(1)=1\, U_n(1)=n+1\, P_n(1)=1\, Lead term = 2^n\,
Square of norm, h_n\, \left\{
\begin{matrix}
\pi   &:~n=0 \\
\pi/2 &:~n\ne 0
\end{matrix}\right.
\pi/2\, \frac{2}{2n+1} 2^n\,n!\,\sqrt{\pi}
Leading term, k_n\, 2^{n-1}\, 2^n\, \frac{(2n)!}{2^n\,(n!)^2}\, 2^n\,
Second term, k'_n\, 0\, 0\, 0\, 0\,
Q\, 1-x^2\, 1-x^2\, 1-x^2\, 1\,
L\, -x\, -3x\, -2x\, -2x\,
R(x) =e^{\int \frac{L(x)}{Q(x)}\,dx} (1-x^2)^{1/2}\, (1-x^2)^{3/2}\, 1-x^2\, e^{-x^2}\,
Constant in diff. equation, {\lambda}_n\, n^2\, n(n+2)\, n(n+1)\, 2n\,
Constant in Rodrigues' formula, e_n\, (-2)^n\,\frac{\Gamma(n+1/2)}{\sqrt{\pi}}\, 2(-2)^n\,\frac{\Gamma(n+3/2)}{(n+1)\,\sqrt{\pi}}\, (-2)^n\,n!\, (-1)^n\,
Recurrence relation, a_n\, 2\, 2\, \frac{2n+1}{n+1}\, 2\,
Recurrence relation, b_n\, 0\, 0\, 0\, 0\,
Recurrence relation, c_n\, 1\, 1\, \frac{n}{n+1}\, 2n\,
Name, and conventional symbol Associated Laguerre, L_n^{(\alpha)} Laguerre, \ L_n
Limits of orthogonality 0, \infty\, 0, \infty\,
Weight, W(x)\, x^{\alpha}e^{-x}\, e^{-x}\,
Standardization Lead term = \frac{(-1)^n}{n!}\, Lead term = \frac{(-1)^n}{n!}\,
Square of norm, h_n\, \frac{\Gamma(n+\alpha+1)}{n!}\, 1\,
Leading term, k_n\, \frac{(-1)^n}{n!}\, \frac{(-1)^n}{n!}\,
Second term, k'_n\, \frac{(-1)^{n+1}(n+\alpha)}{(n-1)!}\, \frac{(-1)^{n+1}n}{(n-1)!}\,
Q\, x\, x\,
L\, \alpha+1-x\, 1-x\,
R(x) =e^{\int \frac{L(x)}{Q(x)}\,dx} x^{\alpha+1}\,e^{-x}\, x\,e^{-x}\,
Constant in diff. equation, {\lambda}_n\, n\, n\,
Constant in Rodrigues' formula, e_n\, n!\, n!\,
Recurrence relation, a_n\, \frac{-1}{n+1}\, \frac{-1}{n+1}\,
Recurrence relation, b_n\, \frac{2n+1+\alpha}{n+1}\, \frac{2n+1}{n+1}\,
Recurrence relation, c_n\, \frac{n+\alpha}{n+1}\, \frac{n}{n+1}\,
Name, and conventional symbol Gegenbauer, C_n^{(\alpha)} Jacobi, P_n^{(\alpha, \beta)}
Limits of orthogonality -1, 1\, -1, 1\,
Weight, W(x)\, (1-x^2)^{\alpha-1/2}\, (1-x)^\alpha(1+x)^\beta\,
Standardization C_n^{(\alpha)}(1)=\frac{\Gamma(n+2\alpha)}{n!\,\Gamma(2\alpha)}\, if \alpha\ne0 P_n^{(\alpha, \beta)}(1)=\frac{\Gamma(n+1+\alpha)}{n!\,\Gamma(1+\alpha)}\,
Square of norm, h_n\, \frac{\pi\,2^{1-2\alpha}\Gamma(n+2\alpha)}{n!(n+\alpha)(\Gamma(\alpha))^2} \frac{2^{\alpha+\beta+1}\,\Gamma(n\!+\!\alpha\!+\!1)\,\Gamma(n\!+\!\beta\!+\!1)}
{n!(2n\!+\!\alpha\!+\!\beta\!+\!1)\Gamma(n\!+\!\alpha\!+\!\beta\!+\!1)}
Leading term, k_n\, \frac{\Gamma(2n+2\alpha)\Gamma(1/2+\alpha)}{n!\,2^n\,\Gamma(2\alpha)\Gamma(n+1/2+\alpha)}\, \frac{\Gamma(2n+1+\alpha+\beta)}{n!\,2^n\,\Gamma(n+1+\alpha+\beta)}\,
Second term, k'_n\, 0\, \frac{(\alpha-\beta)\,\Gamma(2n+\alpha+\beta)}{(n-1)!\,2^n\,\Gamma(n+1+\alpha+\beta)}\,
Q\, 1-x^2\, 1-x^2\,
L\, -(2\alpha+1)\,x\, \beta-\alpha-(\alpha+\beta+2)\,x\,
R(x) =e^{\int \frac{L(x)}{Q(x)}\,dx} (1-x^2)^{\alpha+1/2}\, (1-x)^{\alpha+1}(1+x)^{\beta+1}\,
Constant in diff. equation, {\lambda}_n\, n(n+2\alpha)\, n(n+1+\alpha+\beta)\,
Constant in Rodrigues' formula, e_n\, \frac{(-2)^n\,n!\,\Gamma(2\alpha)\,\Gamma(n\!+\!1/2\!+\!\alpha)}
{\Gamma(n\!+\!2\alpha)\Gamma(\alpha\!+\!1/2)} (-2)^n\,n!\,
Recurrence relation, a_n\, \frac{2(n+\alpha)}{n+1}\, \frac{(2n+1+\alpha+\beta)(2n+2+\alpha+\beta)}{2(n+1)(n+1+\alpha+\beta)}
Recurrence relation, b_n\, 0\, \frac{({\alpha}^2-{\beta}^2)(2n+1+\alpha+\beta)}{2(n+1)(2n+\alpha+\beta)(n+1+\alpha+\beta)}
Recurrence relation, c_n\, \frac{n+2{\alpha}-1}{n+1}\, \frac{(n+\alpha)(n+\beta)(2n+2+\alpha+\beta)}{(n+1)(n+1+\alpha+\beta)(2n+\alpha+\beta)}

See also

Notes

  1. [1], Totik, V., Orthogonal Polynomials, Surveys in Approximation Theory 1(2005) 70–125
  2. J. J. Foncannon, Review of Classical and Quantum Orthogonal Polynomials in One Variable by Mourad Ismail, Mathematical Intelligencer, volume 30, number 1, Winter 2008, pages 54–60.

References

Further reading